Goto

Collaborating Authors

 Islamabad Capital Territory


Do Taliban's drone attacks expose a chink in Pakistan's armour?

Al Jazeera

Do Taliban's drone attacks expose a chink in Pakistan's armour? On the evening of March 13, drones struck three locations across Pakistan. Two children were wounded in Quetta. Civilians were also injured in Kohat and in Rawalpindi, the garrison city that houses the headquarters of Pakistan's armed forces and neighbours the capital, Islamabad. Pakistan's military said the drones were intercepted before reaching their targets.


India's outsourcing industry is worth 300bn. Can it survive AI?

BBC News

India's outsourcing industry is worth $300bn. Indian technology stocks have seen an unprecedented rout over the past few weeks over fears of artificial intelligence upending the traditional outsourcing model that powers the country's $300bn (£223bn) back-office industry. The sell-off - part of a global correction in traditional software and IT stocks - preceded the market nervousness caused by recent geopolitical uncertainty, and is particularly significant for India. Over the past three-and-a-half decades, India's software industry has created millions of white-collar jobs, spawning a new middle class driven by high ambition and strong purchasing power. This, in turn, has fuelled demand for apartments, cars and restaurants across top-tier cities such as Bengaluru, Hyderabad and Gurugram over the past 30 years.





1457c0d6bfcb4967418bfb8ac142f64a-Supplemental.pdf

Neural Information Processing Systems

Reversed Words and Anagrams: Recall that these tasks are of the form "alaok =100 koala". Due to the short length of these tasks, we used 2-grams for filtering (ignoring101 punctuation).



Blind Ultrasound Image Enhancement via Self-Supervised Physics-Guided Degradation Modeling

Khan, Shujaat, Atif, Syed Muhammad, Huh, Jaeyoung, Azhar, Syed Saad

arXiv.org Machine Learning

Ultrasound (US) interpretation is hampered by multiplicative speckle, acquisition blur from the point-spread function (PSF), and scanner- and operator-dependent artifacts. Supervised enhancement methods assume access to clean targets or known degradations; conditions rarely met in practice. We present a blind, self-supervised enhancement framework that jointly deconvolves and denoises B-mode images using a Swin Convolutional U-Net trained with a \emph{physics-guided} degradation model. From each training frame, we extract rotated/cropped patches and synthesize inputs by (i) convolving with a Gaussian PSF surrogate and (ii) injecting noise via either spatial additive Gaussian noise or complex Fourier-domain perturbations that emulate phase/magnitude distortions. For US scans, clean-like targets are obtained via non-local low-rank (NLLR) denoising, removing the need for ground truth; for natural images, the originals serve as targets. Trained and validated on UDIAT~B, JNU-IFM, and XPIE Set-P, and evaluated additionally on a 700-image PSFHS test set, the method achieves the highest PSNR/SSIM across Gaussian and speckle noise levels, with margins that widen under stronger corruption. Relative to MSANN, Restormer, and DnCNN, it typically preserves an extra $\sim$1--4\,dB PSNR and 0.05--0.15 SSIM in heavy Gaussian noise, and $\sim$2--5\,dB PSNR and 0.05--0.20 SSIM under severe speckle. Controlled PSF studies show reduced FWHM and higher peak gradients, evidence of resolution recovery without edge erosion. Used as a plug-and-play preprocessor, it consistently boosts Dice for fetal head and pubic symphysis segmentation. Overall, the approach offers a practical, assumption-light path to robust US enhancement that generalizes across datasets, scanners, and degradation types.


Chameleon: Adaptive Adversarial Agents for Scaling-Based Visual Prompt Injection in Multimodal AI Systems

Zeeshan, M, Satti, Saud

arXiv.org Artificial Intelligence

Multimodal Artificial Intelligence (AI) systems, particularly Vision-Language Models (VLMs), have become integral to critical applications ranging from autonomous decision-making to automated document processing. As these systems scale, they rely heavily on preprocessing pipelines to handle diverse inputs efficiently. However, this dependency on standard preprocessing operations, specifically image downscaling, creates a significant yet often overlooked security vulnerability. While intended for computational optimization, scaling algorithms can be exploited to conceal malicious visual prompts that are invisible to human observers but become active semantic instructions once processed by the model. Current adversarial strategies remain largely static, failing to account for the dynamic nature of modern agentic workflows. To address this gap, we propose Chameleon, a novel, adaptive adversarial framework designed to expose and exploit scaling vulnerabilities in production VLMs. Unlike traditional static attacks, Chameleon employs an iterative, agent-based optimization mechanism that dynamically refines image perturbations based on the target model's real-time feedback. This allows the framework to craft highly robust adversarial examples that survive standard downscaling operations to hijack downstream execution. We evaluate Chameleon against Gemini 2.5 Flash model. Our experiments demonstrate that Chameleon achieves an Attack Success Rate (ASR) of 84.5% across varying scaling factors, significantly outperforming static baseline attacks which average only 32.1%. Furthermore, we show that these attacks effectively compromise agentic pipelines, reducing decision-making accuracy by over 45% in multi-step tasks. Finally, we discuss the implications of these vulnerabilities and propose multi-scale consistency checks as a necessary defense mechanism.


Optimally Deep Networks -- Adapting Model Depth to Datasets for Superior Efficiency

Tareen, Shaharyar Ahmed Khan, Tareen, Filza Khan

arXiv.org Artificial Intelligence

Deep neural networks (DNNs) have provided brilliant performance across various tasks. However, this success often comes at the cost of unnecessarily large model sizes, high computational demands, and substantial memory footprints. Typically, powerful architectures are trained at full depths but not all datasets or tasks require such high model capacity. Training big and deep architectures on relatively low-complexity datasets frequently leads to wasted computation, unnecessary energy consumption, and excessive memory usage, which in turn makes deployment of models on resource-constrained devices impractical. To address this problem, we introduce the concept of Optimally Deep Networks (ODNs), which provides a balance between model depth and task complexity. Specifically, we propose a NAS like training strategy called progressive depth expansion, which begins by training neural networks at shallower depths and incrementally increases their depth as the earlier blocks converge, continuing this process until the target accuracy is reached. ODNs use only the optimal depth for the tasks at hand, removing redundant layers. This cuts down future training and inference costs, lowers the model memory footprint, enhances computational efficiency, and facilitates deployment on edge devices. Empirical results show that the optimal depths of ResNet-18 and ResNet-34 for MNIST and SVHN, achieve up to 98.64 % and 96.44 % reduction in memory footprint, while maintaining a competitive accuracy of 99.31 % and 96.08 %, respectively.